Multimodal Emotion Recognition


Multimodal emotion recognition is the process of recognizing emotions from multiple modalities, such as speech, text, and facial expressions.

The Emotional Baby Is Truly Deadly: Does your Multimodal Large Reasoning Model Have Emotional Flattery towards Humans?

Add code
Aug 06, 2025
Viaarxiv icon

Speech Tokenizer is Key to Consistent Representation

Add code
Jul 09, 2025
Viaarxiv icon

Team RAS in 9th ABAW Competition: Multimodal Compound Expression Recognition Approach

Add code
Jul 02, 2025
Viaarxiv icon

Facial Emotion Learning with Text-Guided Multiview Fusion via Vision-Language Model for 3D/4D Facial Expression Recognition

Add code
Jul 02, 2025
Viaarxiv icon

GSDNet: Revisiting Incomplete Multimodal-Diffusion from Graph Spectrum Perspective for Conversation Emotion Recognition

Add code
Jun 14, 2025
Viaarxiv icon

Towards Robust Multimodal Emotion Recognition under Missing Modalities and Distribution Shifts

Add code
Jun 12, 2025
Viaarxiv icon

MEDUSA: A Multimodal Deep Fusion Multi-Stage Training Framework for Speech Emotion Recognition in Naturalistic Conditions

Add code
Jun 11, 2025
Viaarxiv icon

Multimodal Emotion Recognition in Conversations: A Survey of Methods, Trends, Challenges and Prospects

Add code
May 26, 2025
Viaarxiv icon

MEGC2025: Micro-Expression Grand Challenge on Spot Then Recognize and Visual Question Answering

Add code
Jun 18, 2025
Viaarxiv icon

RoHyDR: Robust Hybrid Diffusion Recovery for Incomplete Multimodal Emotion Recognition

Add code
May 23, 2025
Viaarxiv icon